Implementation and Optimization of Differentiable Neural Computers
نویسنده
چکیده
We implemented and optimized Differentiable Neural Computers (DNCs) as described in the Oct. 2016 DNC paper [1] on the bAbI dataset [25] and on copy tasks that were described in the Neural Turning Machine paper [12]. This paper will give the reader a better understanding of this new and promising architecture through the documentation of the approach in our DNC implementation and our experience of the challenges of optimizing DNCs. Given how recently the DNC paper has come out, other than the original paper, there were no such explanation, implementation and experimentation of the same level of detail as this project has produced, which is why this project will be useful for others who want to experiment with DNCs since this project successfully trained a high performing DNC on the copy task while our DNC performance on the bAbI dataset was better than or equal to our LSTM baseline.
منابع مشابه
A Hybrid Optimization Algorithm for Learning Deep Models
Deep learning is one of the subsets of machine learning that is widely used in Artificial Intelligence (AI) field such as natural language processing and machine vision. The learning algorithms require optimization in multiple aspects. Generally, model-based inferences need to solve an optimized problem. In deep learning, the most important problem that can be solved by optimization is neural n...
متن کاملA Hybrid Optimization Algorithm for Learning Deep Models
Deep learning is one of the subsets of machine learning that is widely used in Artificial Intelligence (AI) field such as natural language processing and machine vision. The learning algorithms require optimization in multiple aspects. Generally, model-based inferences need to solve an optimized problem. In deep learning, the most important problem that can be solved by optimization is neural n...
متن کاملAn efficient one-layer recurrent neural network for solving a class of nonsmooth optimization problems
Constrained optimization problems have a wide range of applications in science, economics, and engineering. In this paper, a neural network model is proposed to solve a class of nonsmooth constrained optimization problems with a nonsmooth convex objective function subject to nonlinear inequality and affine equality constraints. It is a one-layer non-penalty recurrent neural network based on the...
متن کاملNeural Network Implementation in SAS
The estimation or training methods in the neural network literature are usually some simple form of gradient descent algorithm suitable for implementation in hardware using massively parallel computations. For ordinary computers that are not massively parallel, optimization algorithms such as those in several SAS procedures are usually far more efficient. This talk shows how to fit neural netwo...
متن کاملRegularity Conditions for Non-Differentiable Infinite Programming Problems using Michel-Penot Subdifferential
In this paper we study optimization problems with infinite many inequality constraints on a Banach space where the objective function and the binding constraints are locally Lipschitz. Necessary optimality conditions and regularity conditions are given. Our approach are based on the Michel-Penot subdifferential.
متن کامل